Combining Nearest Neighbor Classi ers Through Multiple
نویسنده
چکیده
Combining multiple classiiers is an eeective technique for improving accuracy. There are many general combining algorithms, such as Bagging or Error Correcting Output Coding, that signiicantly improve classiiers like decision trees, rule learners, or neural networks. Unfortunately, many combining methods do not improve the nearest neighbor classiier. In this paper, we present MFS, a combining algorithm designed to improve the accuracy of the nearest neighbor (NN) classiier. MFS combines multiple NN classiiers each using only a random subset of features. The experimental results are encouraging: On 25 datasets from the UCI Repository, MFS sig-niicantly improved upon the NN, k nearest neighbor (kNN), and NN classiiers with forward and backward selection of features. MFS was also robust to corruption by irrelevant features compared to the kNN classiier. Finally, we show that MFS is able to reduce both bias and variance components of error.
منابع مشابه
Combining Nearest Neighbor Classifiers Through Multiple Feature Subsets
Combining multiple classi ers is an e ective technique for improving accuracy. There are many general combining algorithms, such as Bagging or Error Correcting Output Coding, that signi cantly improve classi ers like decision trees, rule learners, or neural networks. Unfortunately, many combining methods do not improve the nearest neighbor classi er. In this paper, we present MFS, a combining a...
متن کاملNearest Neighbor Classi cation from Multiple Feature Subsets
Combining multiple classi ers is an e ective technique for improving accuracy. There are many general combining algorithms, such as Bagging, Boosting, or Error Correcting Output Coding, that signi cantly improve classi ers like decision trees, rule learners, or neural networks. Unfortunately, these combining methods do not improve the nearest neighbor classi er. In this paper, we present MFS, a...
متن کاملEective supra-classi®ers for knowledge base construction
We explore the use of the supra-classi®er framework in the construction of a classi®er knowledge base. Previously, we introduced this framework within which labels produced by old classi®ers are used to improve the generalization performance of a new classi®er for a dierent but related classi®cation task (Bollacker and Ghosh, 1998). We showed empirically that a simple Hamming nearest neighbor ...
متن کاملNearest neighbor classi®er: Simultaneous editing and feature selection
Nearest neighbor classi®ers demand signi®cant computational resources (time and memory). Editing of the reference set and feature selection are two dierent approaches to this problem. Here we encode the two approaches within the same genetic algorithm (GA) and simultaneously select features and reference cases. Two data sets were used: the SATIMAGE data and a generated data set. The GA was fou...
متن کاملCloud Classi cation Using Error-Correcting Output Codes
Novel arti cial intelligence methods are used to classify 16x16 pixel regions (obtained from Advanced Very High Resolution Radiometer (AVHRR) images) in terms of cloud type (e.g., stratus, cumulus, etc.). We previously reported that intelligent feature selection methods, combined with nearest neighbor classi ers, can dramatically improve classi cation accuracy on this task. Our subsequent analy...
متن کامل